Parsimonious Relevance and Concept Models
نویسندگان
چکیده
We describe our participation in the CLEF 2008 Domain Specific track. The research questions we address are threefold: (i) what are the effects of estimating and applying relevance models to the domain specific collection used at CLEF 2008, (ii) what are the results of parsimonizing these relevance models, and (iii) what are the results of applying concept models for blind relevance feedback? Parsimonization is a technique by which the term probabilities in a language model may be re-estimated based on a comparison with a reference model, making the resulting model more sparse and to the point. Concept models are term distributions over vocabulary terms, based on the language associated with concepts in a thesaurus or ontology and are estimated using the documents which are annotated with concepts. Concept models may be used for blind relevance feedback, by first translating a query to concepts and then back to query terms. We find that applying relevance models helps significantly for the current test collection, in terms of both mean average precision and early precision. Moreover, parsimonizing the relevance models helps mean average precision on title-only queries and early precision on title+narrative queries. Our concept models are able to significantly outperform a baseline query-likelihood run, both in terms of mean average precision and early precision on both title-only and title+narrative queries.
منابع مشابه
Parsimonious Relevance Models for Multiple Corpora
We describe a method for applying parsimonious language models to re-estimate the term probabilities assigned by relevance models. We apply our method to six topic sets from test collections in five different genres. Our parsimonious relevance models (i) improve retrieval effectiveness in terms of MAP on all collections, (ii) significantly outperform their non-parsimonious counterparts on most ...
متن کاملThe University of Amsterdam at the CLEF 2008 Domain Specific Track: Parsimonious Relevance and Concept Models
We describe our participation in the CLEF 2008 Domain Specific track. The research questions we address are threefold: (i) what are the effects of estimating and applying relevance models to the domain specific collection used at CLEF 2008, (ii) what are the results of parsimonizing these relevance models, and (iii) what are the results of applying concept models for blind relevance feedback? P...
متن کاملParsimonious Language Models for a Terabyte of Text
The aims of this paper are twofold. Our first aim is to compare results of the earlier Terabyte tracks to the Million Query track. We submitted a number of runs using different document representations (such as full-text, title-fields, or incoming anchor-texts) to increase pool diversity. The initial results show broad agreement in system rankings over various measures on topic sets judged at b...
متن کاملParsimonious Function Representation and Optimal Predictive Model Selection
This paper proposes an intuitively appealing approach to function approximation that yields both parsimonious functional representations and optimal predictive models. Along the lines of the median probability model, the concept of prevalence is introduced and defined in terms of the posterior model probabilities. The posterior distribution of model size is used as the main device to determine ...
متن کاملA Formal Analysis of Relevance
We investigate the notion of relevance as it pertains to ‘commonsense’, subjunctive conditionals. Relevance is taken here as a relation between a property (such as having a broken wing) and a conditional (such as birds typically fly). Specifically, we explore a notion of ‘causative’ relevance, distinct from ‘evidential’ relevance found, for example, in probabilistic approaches. A series of post...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017